Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Recommendation method based on knowledge‑awareness and cross-level contrastive learning
Jie GUO, Jiayu LIN, Zuhong LIANG, Xiaobo LUO, Haitao SUN
Journal of Computer Applications    2024, 44 (4): 1121-1127.   DOI: 10.11772/j.issn.1001-9081.2023050613
Abstract91)   HTML0)    PDF (968KB)(51)       Save

As a kind of side information, Knowledge Graph (KG) can effectively improve the recommendation quality of recommendation models, but the existing knowledge-awareness recommendation methods based on Graph Neural Network (GNN) suffer from unbalanced utilization of node information. To address the above problem, a new recommendation method based on Knowledge?awareness and Cross-level Contrastive Learning (KCCL) was proposed. To alleviate the problem of unbalanced node information utilization caused by the sparse interaction data and noisy knowledge graph that deviate from the true representation of inter-node dependencies during information aggregation, a contrastive learning paradigm was introduced into knowledge-awareness recommendation model of GNN. Firstly, the user-item interaction graph and the item knowledge graph were integrated into a heterogeneous graph, and the node representations of users and items were realized by a GNN based on the graph attention mechanism. Secondly, consistent noise was added to the information propagation aggregation layer for data augmentation to obtain node representations of different levels, and the obtained outermost node representation was compared with the innermost node representation for cross-level contrastive learning. Finally, the supervised recommendation task and the contrastive learning assistance task were jointly optimized to obtain the final representation of each node. Experimental results on DBbook2014 and MovieLens-1m datasets show that compared to the second prior contrastive method, the Recall@10 of KCCL is improved by 3.66% and 0.66%, respectively, and the NDCG@10 is improved by 3.57% and 3.29%, respectively, which verifies the effectiveness of KCCL.

Table and Figures | Reference | Related Articles | Metrics
Location control method for generated objects by diffusion model with exciting and pooling attention
Jinsong XU, Ming ZHU, Zhiqiang LI, Shijie GUO
Journal of Computer Applications    2024, 44 (4): 1093-1098.   DOI: 10.11772/j.issn.1001-9081.2023050634
Abstract114)   HTML5)    PDF (2886KB)(48)       Save

Due to the ambiguity of text and the lack of location information in training data, current state-of-the-art diffusion model cannot accurately control the locations of generated objects in the image under the condition of text prompts. To address this issue, a spatial condition of the object’s location range was introduced, and an attention-guided method was proposed based on the strong correlation between the cross-attention map in U-Net and the image spatial layout to control the generation of the attention map, thus controlling the locations of the generated objects. Specifically, based on the Stable Diffusion (SD) model, in the early stage of the generation of the cross-attention map in the U-Net layer, a loss was introduced to stimulate high attention values in the corresponding location range, and reduce the average attention value outside the range. The noise vector in the latent space was optimized step by step in each denoising step to control the generation of the attention map. Experimental results show that the proposed method can significantly control the locations of one or more objects in the generated image, and when generating multiple objects, it can reduce the phenomenon of object omission, redundant object generation, and object fusion.

Table and Figures | Reference | Related Articles | Metrics
Semi-supervised fake job advertisement detection model based on consistency training
Ruiqi WANG, Shujuan JI, Ning CAO, Yajie GUO
Journal of Computer Applications    2023, 43 (9): 2932-2939.   DOI: 10.11772/j.issn.1001-9081.2022081163
Abstract138)   HTML11)    PDF (2191KB)(94)       Save

The flood of fake job advertisements will not only damage the legitimate rights and interests of job seekers but also disrupt the normal employment order, which results in a poor user experience for job seekers. To effectively detect fake job advertisements, an SSC (Semi-Supervised fake job advertisements detection model based on Consistency training) was proposed. Firstly, the consistency regularization term was applied on all the data to improve the performance of the model. Then, supervised loss and unsupervised loss were integrated through joint training to obtain the semi-supervised loss. Finally, the semi-supervised loss was used to optimize the model. Experimental results on two real datasets EMSCAD (EMployment SCam Aegean Dataset) and IMDB (Internet Movie DataBase) show that SSC achieves the best detection performance when the labeled data are only 20, and the accuracy is increased by 2.2 and 2.8 percentage points compared with the existing advanced semi-supervised learning model UDA (Unsupervised Data Augmentation), and is increased by 3.4 and 11.7 percentage points compared with the deep learning model BERT (Bidirectional Encoder Representations from Transformers). At the same time, SSC has good scalability.

Table and Figures | Reference | Related Articles | Metrics
Attribute reduction algorithm based on cluster granulation and divergence among clusters
Yan LI, Bin FAN, Jie GUO
Journal of Computer Applications    2022, 42 (9): 2701-2712.   DOI: 10.11772/j.issn.1001-9081.2021081371
Abstract242)   HTML10)    PDF (3592KB)(73)    PDF(mobile) (654KB)(12)    Save

Attribute reduction is a hot research topic in rough set theory. Most of the algorithms of attribute reduction for continuous data are based on dominance relations or neighborhood relations. However, continuous datasets do not necessarily have dominance relations in attributes. And the attribute reduction algorithms based on neighborhood relations can adjust the granulation degree through neighborhood radius, but it is difficult to unify the radii due to the different dimensions of attributes and the continuous values of radius parameters, resulting in high computational cost of the whole parameter granulation process. To solve this problem, a multi-granularity attribute reduction strategy based on cluster granulation was proposed. Firstly, the similar samples were classified by the clustering method, and the concepts of approximate set, relative positive region and positive region reduction based on clustering were proposed. Secondly, according to JS (Jensen-Shannon) divergence theory, the difference of data distribution of each attribute among clusters was measured, and representative features were selected to distinguish different clusters. Finally, an attribute reduction algorithm was designed using a discernibility matrix. In the proposed algorithm, the attributes were not required to have ordered relations. Different from neighborhood radius, the clustering parameter was discrete, and the dataset was able to be divided into different granulation degrees by adjusting this parameter. Experimental results on UCI and Kent Ridge datasets show that this attribute reduction algorithm can directly deal with continuous data. At the same time, by using this algorithm, the redundant features in the datasets can be removed while maintaining or even improving the classification accuracy by discrete adjustment of the parameters in a small range.

Table and Figures | Reference | Related Articles | Metrics
Feature construction and preliminary analysis of uncertainty for meta-learning
Yan LI, Jie GUO, Bin FAN
Journal of Computer Applications    2022, 42 (2): 343-348.   DOI: 10.11772/j.issn.1001-9081.2021071198
Abstract461)   HTML67)    PDF (483KB)(186)       Save

Meta-learning is the learning process of applying machine learning methods (meta-algorithms) to seek the mapping between features of a problem (meta-features) and relative performance measures of the algorithm, thereby forming the learning process of meta-knowledge. How to construct and extract meta-features is an important research content. Concerning the problem that most of meta-features used in the existing related researches are statistical features of data, uncertainty modeling was proposed and the impact of uncertainty on learning system was studied. Based on inconsistency of data, complexity of boundary, uncertainty of model output, linear capability to be classified, degree of attribute overlap, and uncertainty of feature space, six kinds of uncertainty meta-features were established for data or models. At the same time,the uncertainty size of the learning problem itself was measured from different perspectives, and specific definitions were given. The correlations between these meta-features were analyzed on artificial datasets and real datasets of a large number of classification problems, and multiple classification algorithms such as K-Nearest Neighbor (KNN) were used to conduct a preliminary analysis of the correlation between meta-features and test accuracy. Results show that the average degree of correlation is about 0.8, indicating that these meta-features have a significant impact on learning performance.

Table and Figures | Reference | Related Articles | Metrics
Blind road recognition algorithm based on color and texture information
PENG Yuqing XUE Jie GUO Yongfang
Journal of Computer Applications    2014, 34 (12): 3585-3588.  
Abstract425)      PDF (738KB)(571)       Save

Concerning the problem that existing blind road recognition method has low recognition rate, simplistic handling, and is easily influenced by light, or shadow, an improved blind road recognition method was proposed. According to the color and texture features of blind road, the algorithm used two segmentation methods respectively including color histogram feature threshold segmentation combined with improved region growing segmentation and fuzzy C-means clustering segmentation for gray level co-occurrence matrix feature. And combined with Canny edge detection and Hough transform algorithm, the proposed algorithm made the blind area separated from the pedestrian area and determines the migration direction for the blind. The experimental results show that the proposed algorithm can segment several kinds of blind road more accurately, detect the boundary and direction of blind road and solve the light and shadow problem partly. It can choose the fastest and the most effective segmentation method adoptively, and can be used in a variety of devices, such as electronic guide ones.

Reference | Related Articles | Metrics
Estimation of RFID tags based on 0-1 distribution
QIAN Xiaojie GUO Hongyuan TIAN Yangguang
Journal of Computer Applications    2013, 33 (08): 2128-2131.  
Abstract686)      PDF (622KB)(523)       Save
In the large-scale Radio Frequency Identification (RFID) system, the estimated time of current tag estimation algorithms increase linearly with the increase of tags, and the deviation is large. Regarding these problems, a new estimation algorithm based on 0-1 distribution was proposed. By using the feature of 0-1 distribution, the algorithm set the specific frame length and selected flag to choose the collection of tags which responded to the command of query. In this way, estimation time was reduced to the logarithmic level of tag number and the deviation was reduced through picking numerous average values randomly. Compared with other algorithms, the simulation results show that the proposed algorithm drops deviation at least by 0.9%, and has less fluctuation.
Reference | Related Articles | Metrics
Fireworks simulation based on CUDA particle system
CHEN Xiuliang LIANG Yingjie GUO Fuliang
Journal of Computer Applications    2013, 33 (07): 2059-2062.   DOI: 10.11772/j.issn.1001-9081.2013.07.2059
Abstract1160)      PDF (603KB)(536)       Save
The elementary theory of the particle system coincides with the objective laws of the natural world. As a result, the particle system can be used for fireworks and other complex phenomena simulation. To solve the problem that the simulation of particle system is of huge computation and memory resources consumption, the paper built the basic particle system model based on Compute Unified Device Architecture (CUDA) framework. The storage and movement update of particles in the model were considered. Then the parallel KD-TRIE neighbor particle search algorithm based on CUDA was studied. Finally, the detailed implementation of the fireworks simulation was discussed based on the CUDA particle system. The results show that the model can simulate the rise and bloom of the fireworks realistically with the frame rate of up to 320 frames per second, enhancing the fidelity and real-time performance of the simulation.
Reference | Related Articles | Metrics
Survey on emerging pattern based contrast mining and applications
DUAN Lei TANG Chang-jie Guozhu DONG YANG Ning GOU Chi
Journal of Computer Applications    2012, 32 (02): 304-308.   DOI: 10.3724/SP.J.1087.2012.00304
Abstract1326)      PDF (945KB)(609)       Save
Contrast mining is one of fairly new hot data mining topics. Contrast mining focuses on knowledge that describes differences between classes and conditions, or describes changes over time. Contrast mining aims at developing techniques to discover patterns or models that contrast, and characterize multiple datasets associated with different classes or conditions. Contrast mining has wide applications in reality, due to its ability of simplifying problems and classifying accurately. Research on the mining and application of emerging patterns represents a major direction of contrast mining. This paper provided a survey of such issue. More specifically, after introducing the background, basic concepts and principles of emerging patterns, the paper analyzed the mining methods of emerging patterns, discussed extended definitions of emerging patterns and their mining, stated methods for constructing emerging pattern based classifiers, and illustrated applications of emerging pattern in several real-world fields. Finally, this paper gave out some topics for future research on emerging pattern based contrast mining.
Reference | Related Articles | Metrics
Filtering of ground point cloud based on scanning line and self-adaptive angle-limitation algorithm
Jie GUO Jian-yong LIU You-liang ZHANG Yu ZHU
Journal of Computer Applications    2011, 31 (08): 2243-2245.   DOI: 10.3724/SP.J.1087.2011.02243
Abstract1513)      PDF (451KB)(876)       Save
Concerning the filtering problem of trees, buildings or other ground objects in field terrain reverse engineering, the disadvantages of conventional angle-limitation algorithm were analyzed, which accumulated errors or used a single threshold and could not meet the requirement of wavy terrain. Therefore, a self-adaptive angle-limitation algorithm based on scanning line was put forward. This method worked through limiting the angle of scanning center, reference point (known ground point) and the point to be sorted, which was adaptive with the wavy terrain. Then the modified point cloud was optimized with a curve fitting method by moving window. The experimental results prove that, the proposed algorithm has a sound control of the macro-terrain, and it can filter the wavy terrain point cloud much better.
Reference | Related Articles | Metrics